Distillation Sparsity Training Algorithm for Accelerating Convolutional Neural Networks in Embedded Systems
نویسندگان
چکیده
The rapid development of neural networks has come at the cost increased computational complexity. Neural are both computationally intensive and memory intensive; as such, minimal energy computing power satellites pose a challenge for automatic target recognition (ATR). Knowledge distillation (KD) can distill knowledge from cumbersome teacher network to lightweight student network, transferring essential information learned by network. Thus, concept KD be used improve accuracy networks. Even when learning there is still redundancy in Traditional fix structure before training, such that training does not situation. This paper proposes sparsity (DST) algorithm based on pruning address above limitations. We first through KD, then pruning, allowing learn which connections essential. DST allows teach pruned directly. proposed was tested CIFAR-100, MSTAR, FUSAR-Ship data sets, with 50% setting. First, new loss function teacher-pruned proposed, showed performance close Second, model (uniformity half-pruning UHP) designed solve problem unstructured facilitate implementation general-purpose hardware acceleration storage. Compared traditional UHP double speed
منابع مشابه
Accelerating Convolutional Neural Network Systems
Convolutional Neural Networks have recently been shown to be highly effective classifiers for image and speech data. Due to the large volume of data required to build useful models, and the complexity of the models themselves, efficiency has become one of the primary concerns. This work shows that frequency domain methods can be utilised to accelerate the performance training, inference, and sl...
متن کاملTowards dropout training for convolutional neural networks
Recently, dropout has seen increasing use in deep learning. For deep convolutional neural networks, dropout is known to work well in fully-connected layers. However, its effect in convolutional and pooling layers is still not clear. This paper demonstrates that max-pooling dropout is equivalent to randomly picking activation based on a multinomial distribution at training time. In light of this...
متن کاملThe Power of Sparsity in Convolutional Neural Networks
Deep convolutional networks are well-known for their high computational and memory demands. Given limited resources, how does one design a network that balances its size, training time, and prediction accuracy? A surprisingly effective approach to trade accuracy for size and speed is to simply reduce the number of channels in each convolutional layer by a fixed fraction and retrain the network....
متن کاملSemi-Supervised Training of Convolutional Neural Networks
In this paper we discuss a method for semi-supervised training of CNNs. By using auto-encoders to extract features from unlabeled images, we can train CNNs to accurately classify images with only a small set of labeled images. We show our method’s results on a shallow CNN using the CIFAR-10 dataset, and some preliminary results on a VGG-16 network using the STL-10 dataset.
متن کاملImplementation of Training Convolutional Neural Networks
Deep learning refers to a shining branch of machine learning that is based on learning levels of representations. Convolutional Neural Networks (CNN) is one kind of deep neural network. It can study concurrently. In this article, we use convolutional neural network to implement the typical face recognition problem which can overcome the influence of pose or resolution in face recognition. Then,...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Remote Sensing
سال: 2023
ISSN: ['2315-4632', '2315-4675']
DOI: https://doi.org/10.3390/rs15102609